Session: Whose Similarity Is It Anyway? Persona-Aware Evaluation for Visual Search
Visual similarity is inherently subjective: a sneaker collector, casual shopper, and teenager may interpret the same image pair in very different ways. Yet, most visual search systems rely on one-size-fits-all evaluation, overlooking the nuances of user expertise, purpose, and aesthetic preference. This talk presents a persona-aware evaluation framework that leverages Large Language Models (LLMs) to simulate user-specific perspectives. We design four distinct personas—Fashion Expert, General Consumer, Tech Evaluator, and Teen User—each assigning relevance scores (1–4) to query–result image pairs, accompanied by commentary. By framing prompts through persona-driven bias, our approach moves beyond generic classification, yielding diverse training labels from a single image pair and uncovering subjectivity-aware failure cases. The framework ensures repeatability, scalability, and interpretability for a task traditionally viewed as subjective. Applied to domains like collectibles, fashion, and electronics, it provides a pathway toward user-segment–aware evaluation and ultimately enables personalization in multimodal visual search systems.
Bio
Shubhangi Tandon is Manager of Applied Research at eBay, where she leads the Content Understanding team driving innovations in Visual Shopping, Generative AI, and multimodal retrieval. She has over ten years of experience advancing large-scale search and recommendation systems, including building eBay’s first embedding-based retrieval models and launching GenAI-powered search and selling experiences that reach millions of users.
Her work spans computer vision, natural language processing, and multimodal modeling, with multiple patents in visual search, retrieval, and navigation. She has published at top venues including IEEE Big Data and SIGDIAL, and holds an M.S. in Computer Science from the University of California, Santa Cruz, and a B.E. in Information Technology from Delhi College of Engineering.